Variational Inference of Correspondence-LDA with Multinomial Generative Process

ثبت نشده
چکیده

In original Correspondence LDA (Corr-LDA) [1] the image region is generated by a multivariate Gaussian distribution. We replace this Gaussian distribution by a multinomial distribution, which gives a modified algorithm for parameter estimation. 1 Corr-LDA with Multinomial Generative Process To keep the consistence, we denote visual word and auditory word as sensory word. Thus, A document (a captioned image/audio) consists of a bag of sensory words and a bag of textual words. Our modified Corr-LDA gives a generative process as follows: 1. Draw topic proportions θ ∼ Dirichlet(α). 2. For each sensory word vm,m ∈ {1, · · · ,M}: (a) Draw topic assignment zm|θ ∼ Multinomial(θ). (b) Draw sensory word vm|zm ∼ Multinomial(πzm) 3. For each textual word wn, n ∈ {1, · · · , N} (a) Draw discrete indexing variable yn ∼ Uniform(1, · · · ,M) (b) Draw textual word wn ∼ Multinomial(βzyn ) Observe that step 2.b is differed from original Corr-LDA paper [1], where D.Blei used a multivariate Gaussian to describe the generative process of sensory word v. The graphical representation of our modified Corr-LDA is depicted in Figure 1. The notations used in this paper are summarized in Table 1.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

TR 09 - 002 Mixed - Membership Naive Bayes Models

In recent years, mixture models have found widespread usage in discovering latent cluster structure from data. A popular special case of finite mixture models are naive Bayes models, where the probability of a feature vector factorizes over the features for any given component of the mixture. Despite their popularity, naive Bayes models suffer from two important restrictions: first, they do not...

متن کامل

The Author-Topic Model for Authors and Documents

We introduce the author-topic model, a generative model for documents that extends Latent Dirichlet Allocation (LDA; Blei, Ng, & Jordan, 2003) to include authorship information. Each author is associated with a multinomial distribution over topics and each topic is associated with a multinomial distribution over words. A document with multiple authors is modeled as a distribution over topics th...

متن کامل

Expectation-Propogation for the Generative Aspect Model

The generative aspect model is an extension of the multinomial model for text that allows word probabilities to vary stochastically across docu­ ments. Previous results with aspect models have been promising, but hindered by the computa­ tional difficulty of carrying out inference and learning. This paper demonstrates that the sim­ ple variational methods of Blei et a!. (200 I) can lead to inac...

متن کامل

Correlated Topic Models

Topic models, such as latent Dirichlet allocation (LDA), have been an effective tool for the statistical analysis of document collections and other discrete data. The LDA model assumes that the words of each document arise from a mixture of topics, each of which is a distribution over the vocabulary. A limitation of LDA is the inability to model topic correlation even though, for example, a doc...

متن کامل

Variational Autoencoders for Collaborative Filtering

We extend variational autoencoders (vaes) to collaborative filtering for implicit feedback. This non-linear probabilistic model enables us to go beyond the limited modeling capacity of linear factor models which still largely dominate collaborative filtering research. We introduce a generative model with multinomial likelihood and use Bayesian inference for parameter estimation. Despite widespr...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2010